338 research outputs found
End-to-end Neural Coreference Resolution
We introduce the first end-to-end coreference resolution model and show that
it significantly outperforms all previous work without using a syntactic parser
or hand-engineered mention detector. The key idea is to directly consider all
spans in a document as potential mentions and learn distributions over possible
antecedents for each. The model computes span embeddings that combine
context-dependent boundary representations with a head-finding attention
mechanism. It is trained to maximize the marginal likelihood of gold antecedent
spans from coreference clusters and is factored to enable aggressive pruning of
potential mentions. Experiments demonstrate state-of-the-art performance, with
a gain of 1.5 F1 on the OntoNotes benchmark and by 3.1 F1 using a 5-model
ensemble, despite the fact that this is the first approach to be successfully
trained with no external resources.Comment: Accepted to EMNLP 201
Generating recommendations for entity-oriented exploratory search
We introduce the task of recommendation set generation for entity-oriented
exploratory search. Given an input search query which is open-ended or
under-specified, the task is to present the user with an easily-understandable
collection of query recommendations, with the goal of facilitating domain
exploration or clarifying user intent. Traditional query recommendation systems
select recommendations by identifying salient keywords in retrieved documents,
or by querying an existing taxonomy or knowledge base for related concepts. In
this work, we build a text-to-text model capable of generating a collection of
recommendations directly, using the language model as a "soft" knowledge base
capable of proposing new concepts not found in an existing taxonomy or set of
retrieved documents. We train the model to generate recommendation sets which
optimize a cost function designed to encourage comprehensiveness,
interestingness, and non-redundancy. In thorough evaluations performed by crowd
workers, we confirm the generalizability of our approach and the high quality
of the generated recommendations
- …